Careers

←Job Openings

Clouder Hadoop Administrator – Baltimore,MD Refer a Person Apply for this Job

Job Description

  • Responsible for planning of Capacity Planning, Infrastructure Planning and version fix to build Hadoop Cluster
  • Capacity Planning of Informatica Big Data Management along with implementation design for grid execution.
  • Interact with the business users, Enterprise Architects and Project Managers to gather the requirements.
  • Designing project Execution Plan along with RACI to capture the milestones and deliver it on time.
  • Installation of Cloudera-Hadoop from scratch with SDLC cycle methodology (Dev, Test, Cert, Production, Disaster Recovery).
  • Implementation of Hadoop Security like Kerberos, Cloudera Key Trustee Server and Key Trustee Management Systems.
  • Enable Sentry for RBAC (role-based access control) to have a privilege level access to the data in HDFS.
  • Enabling data encryption at rest and at motion with TLS/SSL.
  • Optimization and Performance tuning the cluster parameters basing on the outputs of Teragen/Terasort.
  • Perform upgrades to Cloudera Manager, CDH along with support for Linux Server Patching from RHEL 7.1 to 7.4(maipo).
  • Have knowledge on a POC for KUDU, in order to check the integrity into my project use case implementation.
  • Installation of Informatica Big-Data-Management edition (10.1.0) from scratch at development, Test, Certification and Production environments.
  • Upgradation of Informatica from 10.1.0 to 10.1.1HotFix1 and 10.1.1HF1 to 10.2.1.
  • Establishment of connections between Hadoop and Informatica to perform Dynamic Mappings and Hive updates.
  • Enable Informatica as a data ingestion tool for Hadoop by creating and testing the connections from different databases like Mysql, Microsoft Sql,Oracle, Hive, Hdfs, Teradata.
  • Designing and Implementation of Backup and Disaster Recovery strategy based out of Cloudera BDR utility for Batch applications and Kafka mirror maker for real-time streaming applications.
  • Aligning with development and architecture teams to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
  • Enable the consumers to use the Data in Hive Tables from Tableau desktop as part of the requirement.
  • Establish the connection between Teradata Studio Express and Impala, so as to enable the consumer group for an easy migration to Hadoop query engines.
  • Integrate CA-7 enterprise scheduler to run the Jobs in both Hadoop/Informatica.
  • Implement GIT version control basing out of NFS shared drive for Hadoop and also integrated it to the Eclipse IDE.
  • Enable Sub-Version(svn) as version control for Informatica 10.1.1HF1 and 10.2.1.
  • Deploye multiple EBF’s (emergency bug fixes) for Informatica and Hadoop.
  • Should be able to develope kafka into Hadoop to enable real-time streaming applications.
  • Should be able work with delivery teams for User-Provisioning.

Required Skills:

  • A minimum of bachelor's degree in computer science or equivalent.
  • Cloudrea Hadoop(CDH), Cloudera Manager, Informatica Bigdata Edition(BDM), HDFS, Yarn, MapReduce, Hive, Impala, KUDU, Sqoop, Spark, Kafka, HBase, Teradata Studio Express, Teradata, Tableau, Kerberos, Active Directory, Sentry, TLS/SSL, Linux/RHEL, Unix Windows, SBT, Maven, Jenkins, Oracle, MS SQL Server, Shell Scripting, Eclipse IDE, Git, SVN
  • Must have strong problem-solving and analytical skills
  • Must have the ability to identify complex problems and review related information to develop and evaluate options and implement solutions.

If you are interested in working in a fast-paced, challenging, fun, entrepreneurial environment and would like to have the opportunity of being a part of this fascinating industry, Send resumes. to HSTechnologies LLC, 2801 W Parker Road, Suite #5 Plano, TX - 75023 or email your resume to hr@sbhstech.com.